Computably Continuous Reinforcement-Learning Objectives Are PAC-Learnable
نویسندگان
چکیده
In reinforcement learning, the classic objectives of maximizing discounted and finite-horizon cumulative rewards are PAC-learnable: There algorithms that learn a near-optimal policy with high probability using finite amount samples computation. recent years, researchers have introduced corresponding reinforcement-learning beyond rewards, such as specified linear temporal logic formulas. However, questions about PAC-learnability these new remained open. This work demonstrates general through sufficient conditions for in two analysis settings. particular, considers only sample complexity, we prove if an objective given oracle is uniformly continuous, then it PAC-learnable. Further, computational computable, other words, procedure computes successive approximations objective's value, We give three applications our condition on from literature previously unknown Overall, result helps verify existing objectives' PAC-learnability. Also, some studied not continuous been shown to be PAC-learnable, results could guide design PAC-learnable objectives.
منابع مشابه
PAC Reinforcement Learning in Noisy Continuous Worlds
Continuous state spaces and stochastic, switching dynamics characterize a number of rich, real-world domains, such as robot navigation across varying terrain. We describe a reinforcement-learning algorithm for learning in such domains. We prove that this algorithm is provably approximately correct for certain environments. Unfortunately, no optimal planning techniques exist in general for such ...
متن کاملStabiliser states are efficiently PAC-learnable
The exponential scaling of the wave function is a fundamental property of quantum systems with far reaching implications in our ability to process quantum information. A problem where these are particularly relevant is quantum state tomography. State tomography, whose objective is to obtain a full description of a quantum system, can be analysed in the framework of computational learning theory...
متن کاملPAC Associative Reinforcement Learning
General algorithms for the reinforcement learning problem typically learn policies in the form of a table that directly maps the states of the environment into actions. When the state-space is large these methods become impractical. One approach to increase e ciency is to restrict the class of policies by considering only policies that can be described using some xed representation. This paper ...
متن کاملPAC Continuous State Online Multitask Reinforcement Learning with Identification
One key feature of a general intelligent autonomous agent is to be able to learn from past experience to improve future performance. In this paper we consider how an agent can leverage prior experience from performing reinforcement learning in order to learn faster in future tasks. We introduce the first, to our knowledge, probably approximately correct (PAC) RL algorithm COMRLI for sequential ...
متن کاملFirst-Order jk-Clausal Theories are PAC-Learnable
We present positive PAC-learning results for the nonmonotonic inductive logic programming setting. In particular, we show that first order range-restricted clausal theories that consist of clauses with up to k literals of size at most j each are polynomialsample polynomial-time PAC-learnable with one-sided error from positive examples only. In our framework, concepts are clausal theories and ex...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence
سال: 2023
ISSN: ['2159-5399', '2374-3468']
DOI: https://doi.org/10.1609/aaai.v37i9.26273